A Comparative Study of L1-distortion Limited Image Compression Algorithms

نویسندگان

  • Hannes Hartenstein
  • Ralf Herz
  • Dietmar Saupe
چکیده

In L1-distortion limited compression each single pixel value is only changed by maximal grey values. In this paper we present a theoretical framework for L1-distortion limited compression that covers several recently proposed methods. The basics of each of these methods are described. We give a comparison of coding results for the Lenna test image, a coronary angiogram, and a Landsat image. Results are reported for various tolerances. Standard DPCM is used as a reference. While this paper gives an overview over various algorithms, the main purpose is to indicate what level of compression can be expected when limiting the error in L1-distortion sense. 1. A L1-DISTORTION LIMITED COMPRESSION FRAMEWORK In many applications, for example medical imagery, SAR imagery, or numerical weather simulations, the large amount of data to be stored or transmitted asks for data compression. Since lossless coding usually gives a compression ratio of at most 4:1, lossy coding methods have to be employed when higher compression ratios are needed. Most lossy compression schemes operate by minimizing some average error measure such as the root mean square error. However, in error critical applications such as medical imagery or target recognition, such average error measures are inappropriate. Instead, there is usually a need for a guarantee that a single pixel has not been changed by more than a certain tolerance (which may depend on the pixel location). Thus, the error in each pixel has to be controlled. In this paper we consider a L1-distortion limited compression scheme with global tolerance . For such an encoding method the code for a one-dimensional signal s 2 Zn represents a reconstruction signal ŝ 2 N (s), where N (s) := ft 2 Znj kt sk1 g ; kt sk1 = max i2f1::ng jti sij: For = 0 this leads to lossless compression. If is small, the term ' near-lossless coding' appears to be 0 1 2 3 4 5 6 7 8 1 2 3 4 5 6 7 8 9 10 Figure 1: Each left-to-right path is an element of N (s) with s = (3; 3; 2; 1; 3; 5; 7; 6; 5; 4). justified. N (s) can be seen as the set of all left-toright paths in a trellis as depicted in Figure 1. Which of the (2 + 1)n elements of N (s) can be coded most efficiently? All coding methods described below use a lossless coding strategy C and try to determine, at least approximately or heuristically, the element of N (s) that can be coded most efficiently using C. Mathematically, let C be the set of lossless coding methods. For example, C 2 C could be a 0-order entropy coder. Then the coding problem for this particular C is: find t 2 N (s) such that `(C(t )) = min t2N (s) `(C(t)) ( ) where ` ( ) gives the length of the code or some estimate thereof. In the following sections we give short descriptions of several L1-based compression methods. Most of these methods were implemented and tested with the images given in Figure 2. 2. QUANTIZATION VS PRECONDITIONING In the problem formulation of the previous section the signal to be coded can be modified in each component independently of the other components. Thus, for a signal s and i 6= j it is possible that s(i) = s(j) but ŝ(i) 6= ŝ(j). In other words, ŝ is in general not simply the result of using a quantizer q for the values of s. We refer to ŝ as a preconditioned version of s in contrast to a quantized version. The emphasis in this paper is on preconditioning. Nevertheless, the problem of finding the quantization function such that the quantized version of s has minimal 0-order entropy can be solved in O(n) time using dynamic programming [1]. It is also shown that for a tolerance > 0 the entropy savings are at most log2(2 + 1) per pixel. 3. ENTROPY-CODED DPCM BASED METHODS The entropy coding of the prediction residuals of a DPCM scheme is a standard method for lossless compression. It can easily be modified to serve as an L1 distortion based compression method. DPCM1. The signal is uniformly quantized with quantization bin size 2 +1. Thus, a quantized version of the original signal is computed. Then the residuals of a linear predictor are entropy coded. The disadvantage of this method is that for larger there are only a few different grey levels leading to ' plateau effects' . DPCM2. No a priori grey value reduction is performed, but the prediction error of the DPCM scheme is uniformly quantized to match the desired tolerance . When the predictor coefficients are not integer values, this method does not coincide with the method DPCM1 and does not show the plateau effects. Results for several medical images are reported in [2]. In the above mentioned methods, there is actually no mechanism to minimize the entropy of the error sequence. When we use a lossless predictive coder followed by an entropy coder, the optimization problem ( ) asks for the path in the trellis whose corresponding residual sequence has minimum entropy. We conjecture that this optimization problem is NP-hard. Note that the complexity depends on the signal length n and the tolerance . We applied genetic algorithms (GA) [3, 4] to solve this optimization problem for a signal s 2 Zn and a tolerance . GA. In our setting a chromosome c is a word of length n over the alphabet f ; ; 0; ; g and represents the signal s + c 2 N (s). The genetic operations are 2-point crossover and mutation. We use roulette wheel parent selection. The evaluation of a chromosome is given by the entropy of the distribution of prediction residuals of ŝ = s + c. For the fitness function we use exponential ranking. Large tests for the determination of suitable parameters were performed. The results obtained with the GA approach are rather disappointing. For example, as a signal s a line of the image Lenna was taken. The entropy of that signal is 7.0, after quantization with tolerance = 2, and after prediction the sequence can be coded with an entropy of 3.1. The solution found with the GA only gave an entropy of 3.9. Thus, the GA is not even able to beat the method DPCM1. The minimum-entropy constrained-error DPCM (MECE) of [5] is another method that tries to minimizes the entropy of the prediction residual sequence. It uses an iterative optimization method that arrives at a local optimum. MECE. Assume that an ideal entropy coder is given for a fixed residual distribution. To find the optimal element of N (s) for this coder one has to solve a shortest path problem. This can easily be done via Dynamic Programming. Now, using an entropy coder that is optimal for the actual residual distribution will give a decrease in entropy. These two steps are performed iteratively until a stopping criterion is matched. For images a two dimensional 3-tap predictor is used and the images are coded row by row. The results can be further improved by using a 1-order entropy coder with a certain number of contexts. In the above mentioned methods the predictor and the contexts are fixed. Of course, it would be advantageous to include the choice of predictor coefficients and context into the optimization problem; clearly, this makes the problem even more complicated. A sophisticated method that uses adaptive context modeling to correct prediction biases is the L1-constrained CALIC [6]. The converse problem of determining a predictor such that the prediction residuals have minimum entropy was investigated for lossless coding in [7]. 4. PIECEWISE LINEAR CODING Piecewise linear coding (PCL) is a generalization of Run Length Encoding. It is also called fan-based coding; for an extensive overview see [8]. In piecewise linear coding a signal is split into segments each of which can be described by a linear function. Each segment then is coded by the length of the segment and the slope parameter. The constant additive part of the function is implicitly given by the previous segment; only for the first segment the initial signal value has to be coded. For example, the signal in Figure 1 is represented as 3(1,0)(2,-1)(3,2)(3,-1). In the case that `( ) counts the number of segments the optimization ( ) can be solved in O(n2)-time via Dynamic Programming [9]. In [10, 11] a suboptimal greedy method that works in linear time is proposed for the same optimization problem. Essentially, it works as follows. The image is transformed into a 1-dimensional signal, e.g., by a Hilbert-Peano scan. Then the linear segments are successively determined: starting at the endpoint of the last determined segment, the new segment is chosen to be the one of greatest possible length. Finally, an 0-order entropy coder is applied to the list of segment lengths and segment slopes. Better results can be obtained when the length of the 0-order entropy code is minimized in place of Figure 2: The 512 512 8-bit test images Lenna, Angio, Landsat. = 0 = 1 = 2 = 3 = 4 = 5 = 10 = 15 bpp bpp rms bpp rms bpp rms bpp rms bpp rms bpp rms bpp rms DPCM1 4.6 3.0 0.8 2.3 1.4 1.9 2.0 1.6 2.6 1.4 3.2 0.9 6.1 0.7 8.9 DPCM2 4.5 3.0 0.8 2.3 1.4 1.9 2.0 1.7 2.6 1.5 3.2 1.0 5.9 0.8 8.7 MECE 4.6 3.1 0.8 2.5 1.4 2.0 1.9 1.7 2.5 1.4 3.0 0.7 5.3 0.5 7.6 PLC 5.0 4.2 0.8 3.5 1.5 3.0 2.2 2.5 2.8 2.2 3.3 1.3 5.9 0.9 8.5 DLVQ 4.7 3.2 0.8 2.5 1.4 2.0 2.0 1.7 2.5 1.3 2.9 0.7 4.9 0.6 6.8 SPIHT 4.5 3.0 0.8 2.3 1.4 1.9 1.8 1.5 2.1 1.2 2.3 0.6 3.2 0.4 3.9 Table 1: Compression results for the Lenna image (bpp : = bits per pixel, rms : = root mean square error). = 0 = 1 = 2 = 3 = 4 = 5 = 10 = 15 bpp bpp rms bpp rms bpp rms bpp rms bpp rms bpp rms bpp rms DPCM1 3.9 2.5 0.8 1.9 1.4 1.5 1.9 1.3 2.5 1.2 3.1 0.8 6.0 0.6 9.0 DPCM2 3.9 2.6 0.9 2.0 1.4 1.7 2.0 1.5 2.6 1.3 3.1 1.0 6.0 0.9 8.9 MECE 3.9 2.6 0.8 2.0 1.3 1.7 1.9 1.5 2.5 1.2 3.0 0.5 5.2 0.3 7.0 PLC 4.5 3.9 0.8 3.3 1.5 2.8 2.3 2.3 2.9 2.0 3.5 0.9 5.9 0.5 7.9 DLVQ 4.2 2.6 0.8 1.9 1.4 1.5 2.0 1.3 2.5 1.0 3.0 0.5 4.8 0.4 6.5 SPIHT 4.0 2.5 0.8 1.9 1.4 1.6 1.7 1.2 1.9 1.0 2.2 0.3 4.2 0.1 4.9 Table 2: Compression results for the Angio image. = 0 = 1 = 2 = 3 = 4 = 5 = 10 = 15 bpp bpp rms bpp rms bpp rms bpp rms bpp rms bpp rms bpp rms DPCM1 4.8 4.6 0.8 4.1 1.4 3.3 1.9 3.2 2.5 3.0 3.1 2.1 5.8 1.7 8.8 DPCM2 5.7 4.5 0.8 3.8 1.4 3.3 2.0 3.0 2.6 2.7 3.1 1.9 6.0 1.4 8.7 MECE 4.8 4.1 0.7 4.0 1.4 3.5 2.1 3.1 2.7 3.0 3.2 2.0 5.8 1.4 8.7 PLC 4.6 5.1 0.8 5.1 1.4 4.7 2.3 4.6 2.9 4.4 3.5 3.2 7.1 2.3 10.1 DLVQ 6.0 4.4 0.8 3.7 1.4 3.2 2.0 2.9 2.5 2.6 3.1 1.7 6.0 1.2 8.6 SPIHT 5.9 4.6 0.8 3.9 1.4 3.4 2.0 3.0 2.5 2.7 3.1 1.9 5.9 1.4 7.2 Table 3: Compression results for the Landsat image. the number of segments [9]. However, in this case the global optimum can no longer be achieved by linear programming alone. Instead, an iterative procedure similar to that in [5] can be applied leading to a locally optimal solution. 5. DISTORTION-LIMITED VQ The DLVQ method proposed by [12] is a multistage vector quantization approach. The image is partitioned, e.g., in 4 4 blocks. A block is coded by a predictive vector quantizer. If there is a component with a distortion above the threshold , the error residual block is coded with another codebook and a second index has to be sent. If the error tolerance is still not satisfied, a scalar quantizer is applied to the second-order residuals which finally guarantees that the error tolerance criterion is satisfied. The special feature of [12] is a codebook design algorithm that reduces the number of large distortion

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

A Comparative Study of -distortion Limited Image Compression Algorithms

In -distortion limited compression each single pixel value is only changed by maximal grey values. In this paper we present a theoretical framework for -distortion limited compression that covers several recently proposed methods. The basics of each of these methods are described. We give a comparison of coding results for the Lenna test image, a coronary angiogram, and a Landsat image. Results...

متن کامل

Dimensionality reduction based on an overcomplete basis

We discuss a strategy of sparse approximation that is based on the use of an overcomplete basis, and evaluate its performance when a random matrix is used as this basis. A small combination of basis vectors is chosen from a given overcomplete basis, according to a given compression rate, such that they compactly represent the target data with as small a distortion as possible. As a selection me...

متن کامل

فشرده‌سازی تصویر با کمک حذف و کدگذاری هوشمندانه اطلاعات تصویر و بازسازی آن با استفاده از الگوریتم های ترمیم تصویر

Compression can be done by lossy or lossless methods. The lossy methods have been used more widely than the lossless compression. Although, many methods for image compression have been proposed yet, the methods using intelligent skipping proper to the visual models has not been considered in the literature. Image inpainting refers to the application of sophisticated algorithms to replace lost o...

متن کامل

Final Report on Evaluation of Synthetic Aperture Radar (SAR) Image Compression Techniques

With the improvement of synthetic aperture radar (SAR) technology, larger areas are being imaged and the resolution of the images has increased. Larger images have to be transmitted and stored. Due to the limited storage and/or downlink capacity on the airplane or satellite, the volume of the data must be reduced. This makes compression of SAR images with minimal loss of information important. ...

متن کامل

Intelligent scalable image watermarking robust against progressive DWT-based compression using genetic algorithms

Image watermarking refers to the process of embedding an authentication message, called watermark, into the host image to uniquely identify the ownership. In this paper a novel, intelligent, scalable, robust wavelet-based watermarking approach is proposed. The proposed approach employs a genetic algorithm to find nearly optimal positions to insert watermark. The embedding positions coded as chr...

متن کامل

Image-compressors Performance Evaluation Based on Power Measurement

The criteria applied on the choice of the compression algorithm is generally based on distortion, compression rate, processing time and hardware limitation. On most cases, the energy expenditure is not considered, however, in some applications, the energy expenditure is a good parameter for choosing the algorithms (E.g. embedded systems that apply the use of batteries). Thus, this article propo...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 1997